324 research outputs found
Nowhere to Hide: Cross-modal Identity Leakage between Biometrics and Devices
Along with the benefits of Internet of Things (IoT) come potential privacy risks, since billions of the connected devices are granted permission to track information about their users and communicate it to other parties over the Internet. Of particular interest to the adversary is the user identity which constantly plays an important role in launching attacks. While the exposure of a certain type of physical biometrics or device identity is extensively studied, the compound effect of leakage from both sides remains unknown in multi-modal sensing environments. In this work, we explore the feasibility of the compound identity leakage across cyber-physical spaces and unveil that co-located smart device IDs (e.g., smartphone MAC addresses) and physical biometrics (e.g., facial/vocal samples) are side channels to each other. It is demonstrated that our method is robust to various observation noise in the wild and an attacker can comprehensively profile victims in multi-dimension with nearly zero analysis effort. Two real-world experiments on different biometrics and device IDs show that the presented approach can compromise more than 70\% of device IDs and harvests multiple biometric clusters with ~94% purity at the same time
When NAS Meets Watermarking: Ownership Verification of DNN Models via Cache Side Channels
We present a novel watermarking scheme to verify the ownership of DNN models.
Existing solutions embedded watermarks into the model parameters, which were
proven to be removable and detectable by an adversary to invalidate the
protection. In contrast, we propose to implant watermarks into the model
architectures. We design new algorithms based on Neural Architecture Search
(NAS) to generate watermarked architectures, which are unique enough to
represent the ownership, while maintaining high model usability. We further
leverage cache side channels to extract and verify watermarks from the
black-box models at inference. Theoretical analysis and extensive evaluations
show our scheme has negligible impact on the model performance, and exhibits
strong robustness against various model transformations
Biologically Plausible Learning on Neuromorphic Hardware Architectures
With an ever-growing number of parameters defining increasingly complex
networks, Deep Learning has led to several breakthroughs surpassing human
performance. As a result, data movement for these millions of model parameters
causes a growing imbalance known as the memory wall. Neuromorphic computing is
an emerging paradigm that confronts this imbalance by performing computations
directly in analog memories. On the software side, the sequential
Backpropagation algorithm prevents efficient parallelization and thus fast
convergence. A novel method, Direct Feedback Alignment, resolves inherent layer
dependencies by directly passing the error from the output to each layer. At
the intersection of hardware/software co-design, there is a demand for
developing algorithms that are tolerable to hardware nonidealities. Therefore,
this work explores the interrelationship of implementing bio-plausible learning
in-situ on neuromorphic hardware, emphasizing energy, area, and latency
constraints. Using the benchmarking framework DNN+NeuroSim, we investigate the
impact of hardware nonidealities and quantization on algorithm performance, as
well as how network topologies and algorithm-level design choices can scale
latency, energy and area consumption of a chip. To the best of our knowledge,
this work is the first to compare the impact of different learning algorithms
on Compute-In-Memory-based hardware and vice versa. The best results achieved
for accuracy remain Backpropagation-based, notably when facing hardware
imperfections. Direct Feedback Alignment, on the other hand, allows for
significant speedup due to parallelization, reducing training time by a factor
approaching N for N-layered networks
Unified Medical Image Pre-training in Language-Guided Common Semantic Space
Vision-Language Pre-training (VLP) has shown the merits of analysing medical
images, by leveraging the semantic congruence between medical images and their
corresponding reports. It efficiently learns visual representations, which in
turn facilitates enhanced analysis and interpretation of intricate imaging
data. However, such observation is predominantly justified on single-modality
data (mostly 2D images like X-rays), adapting VLP to learning unified
representations for medical images in real scenario remains an open challenge.
This arises from medical images often encompass a variety of modalities,
especially modalities with different various number of dimensions (e.g., 3D
images like Computed Tomography). To overcome the aforementioned challenges, we
propose an Unified Medical Image Pre-training framework, namely UniMedI, which
utilizes diagnostic reports as common semantic space to create unified
representations for diverse modalities of medical images (especially for 2D and
3D images). Under the text's guidance, we effectively uncover visual modality
information, identifying the affected areas in 2D X-rays and slices containing
lesion in sophisticated 3D CT scans, ultimately enhancing the consistency
across various medical imaging modalities. To demonstrate the effectiveness and
versatility of UniMedI, we evaluate its performance on both 2D and 3D images
across 10 different datasets, covering a wide range of medical image tasks such
as classification, segmentation, and retrieval. UniMedI has demonstrated
superior performance in downstream tasks, showcasing its effectiveness in
establishing a universal medical visual representation
SenseFi: A library and benchmark on deep-learning-empowered WiFi human sensing
Over the recent years, WiFi sensing has been rapidly developed for privacy-preserving, ubiquitous human-sensing applications, enabled by signal processing and deep-learning methods. However, a comprehensive public benchmark for deep learning in WiFi sensing, similar to that available for visual recognition, does not yet exist. In this article, we review recent progress in topics ranging from WiFi hardware platforms to sensing algorithms and propose a new library with a comprehensive benchmark, SenseFi. On this basis, we evaluate various deep-learning models in terms of distinct sensing tasks, WiFi platforms, recognition accuracy, model size, computational complexity, and feature transferability. Extensive experiments are performed whose results provide valuable insights into model design, learning strategy, and training techniques for real-world applications. In summary, SenseFi is a comprehensive benchmark with an open-source library for deep learning in WiFi sensing research that offers researchers a convenient tool to validate learning-based WiFi-sensing methods on multiple datasets and platforms.Nanyang Technological UniversityPublished versionThis research is supported by NTU Presidential Postdoctoral Fellowship, ââAdaptive Multi-modal Learning for Robust Sensing and Recognition in Smart Citiesââ project fund (020977-00001), at the Nanyang Technological University, Singapore
- âŠ